Optimal recovery of precision matrix for Mahalanobis distance from high-dimensional noisy observations in manifold learning

نویسندگان

چکیده

Abstract Motivated by establishing theoretical foundations for various manifold learning algorithms, we study the problem of Mahalanobis distance (MD) and associated precision matrix estimation from high-dimensional noisy data. By relying on recent transformative results in covariance estimation, demonstrate sensitivity MD to measurement noise, determining exact asymptotic signal-to-noise ratio at which fails, quantifying its performance otherwise. In addition, an appropriate loss function, propose asymptotically optimal shrinker, is shown be beneficial over classical implementation MD, both analytically simulations. The result extended setup, where nonlinear interaction between curvature noise taken care of. developed solution applied a multi-scale reduction dynamical system analysis.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Optimal Tangent Plane Recovery From Noisy Manifold Samples

Constructing an efficient parametrization of a large, noisy data set of points lying close to a smooth manifold in high dimension remains a fundamental problem. One approach consists in recovering a local parametrization using the local tangent plane. Principal component analysis (PCA) is often the tool of choice, as it returns an optimal basis in the case of noise-free samples from a linear su...

متن کامل

Mahalanobis Distance Learning for Person Re-identification

Recently, Mahalanobis metric learning has gained a considerable interest for single-shot person re-identification. The main idea is to build on an existing image representation and to learn a metric that reflects the visual camera-to-camera transitions, allowing for a more powerful classification. The goal of this chapter is twofold. We first review the main ideas of Mahalanobis metric learning...

متن کامل

Optimal Rescaling and the Mahalanobis Distance

One of the basic problems in data analysis lies in choosing the optimal rescaling (change of coordinate system) to study properties of a given data-set Y . The classical Mahalanobis approach has its basis in the classical normalization/rescaling formula Y 3 y → Σ Y · (y − mY ), where mY denotes the mean of Y and ΣY the covariance matrix . Based on the cross-entropy we generalize this approach a...

متن کامل

Sharp thresholds for high-dimensional and noisy recovery of sparsity

The problem of consistently estimating the sparsity pattern of a vector β∗ ∈ R based on observations contaminated by noise arises in various contexts, including subset selection in regression, structure estimation in graphical models, sparse approximation, and signal denoising. We analyze the behavior of l1-constrained quadratic programming (QP), also referred to as the Lasso, for recovering th...

متن کامل

Sample Complexity of Learning Mahalanobis Distance Metrics

Metric learning seeks a transformation of the feature space that enhances prediction quality for a given task. In this work we provide PAC-style sample complexity rates for supervised metric learning. We give matching lowerand upper-bounds showing that sample complexity scales with the representation dimension when no assumptions are made about the underlying data distribution. In addition, by ...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Information and Inference: A Journal of the IMA

سال: 2022

ISSN: ['2049-8772', '2049-8764']

DOI: https://doi.org/10.1093/imaiai/iaac010